Weak-stability and saddle point theorems for a multiobjective optimization problem with an infinite number of constraints
نویسندگان
چکیده
منابع مشابه
Two-Level Optimization Problems with Infinite Number of Convex Lower Level Constraints
This paper proposes a new form of optimization problem which is a two-level programming problem with infinitely many lower level constraints. Firstly, we consider some lower level constraint qualifications (CQs) for this problem. Then, under these CQs, we derive formula for estimating the subdifferential of its valued function. Finally, we present some necessary optimality condit...
متن کاملMultiobjective optimization problem with variational inequality constraints
We study a general multiobjective optimization problem with variational inequality, equality, inequality and abstract constraints. Fritz John type necessary optimality conditions involving Mordukhovich coderivatives are derived. They lead to Kuhn-Tucker type necessary optimality conditions under additional constraint qualifications including the calmness condition, the error bound constraint qu...
متن کاملWEAK AND STRONG DUALITY THEOREMS FOR FUZZY CONIC OPTIMIZATION PROBLEMS
The objective of this paper is to deal with the fuzzy conic program- ming problems. The aim here is to derive weak and strong duality theorems for a general fuzzy conic programming. Toward this end, The convexity-like concept of fuzzy mappings is introduced and then a speci c ordering cone is established based on the parameterized representation of fuzzy numbers. Un- der this setting, duality t...
متن کاملGeneralized iterative methods for solving double saddle point problem
In this paper, we develop some stationary iterative schemes in block forms for solving double saddle point problem. To this end, we first generalize the Jacobi iterative method and study its convergence under certain condition. Moreover, using a relaxation parameter, the weighted version of the Jacobi method together with its convergence analysis are considered. Furthermore, we extend a method...
متن کاملOn the saddle point problem for non-convex optimization
A central challenge to many fields of science and engineering involves minimizing non-convex error functions over continuous, high dimensional spaces. Gradient descent or quasi-Newton methods are almost ubiquitously used to perform such minimizations, and it is often thought that a main source of difficulty for the ability of these local methods to find the global minimum is the proliferation o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: TURKISH JOURNAL OF MATHEMATICS
سال: 2019
ISSN: 1303-6149
DOI: 10.3906/mat-1812-16